Max Margin Dimensionality Reduction

نویسنده

  • Gal Chechik
چکیده

A fundamental problem in machine learning is to extract compact but relevant representations of empirical data. Relevance can be measured by the ability to make good decisions based on the representations, for example in terms of classification accuracy. Compact representations can lead to more human-interpretable models, as well as improve scalability. Furthermore, in multi-class and multi-task problems, learning a unified input representation that is shared among classes or tasks can reduce the number of free parameters and sample-complexity. Sharing representation is also a powerful method for transfer learning [4, 11], where the representation learned from some tasks is used to facilitate learning other tasks. The naive but common practice for finding a compact representation is an unsupervised pre-processing phase (like clustering or PCA) followed by classification in the reduced space. A preferable approach is to learn the features simultaneously with the classifier. Several methods approached this problem, including the well known Fisher’s Linear Discriminant Analysis (LDA), and its variants [7], learning metrics for kNN [9, 8], and extracting features for multi-task classification [3, 2]. Here we focus on finding low-dimensional linear projections that are optimized for support vector machines, in a singleor multitask setting. Formaly, we have k = 1, . . . ,K binary classification tasks, each with labeled data xkik ∈ R , y ik ∈ {−1,+1}, ∀ik = 1, . . . , nk, we look for a rank-d linear projection A, d ≤ D, that is shared across all tasks, and K classifiers w, one for each task. We optimize jointly over all the classifiers w = {

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bayesian Maximum Margin Principal Component Analysis

Supervised dimensionality reduction has shown great advantages in finding predictive subspaces. Previous methods rarely consider the popular maximum margin principle and are prone to overfitting to usually small training data, especially for those under the maximum likelihood framework. In this paper, we present a posterior-regularized Bayesian approach to combine Principal Component Analysis (...

متن کامل

Bayesian Max-margin Multi-Task Learning with Data Augmentation

Both max-margin and Bayesian methods have been extensively studied in multi-task learning, but have rarely been considered together. We present Bayesian max-margin multi-task learning, which conjoins the two schools of methods, thus allowing the discriminative max-margin methods to enjoy the great flexibility of Bayesian methods on incorporating rich prior information as well as performing nonp...

متن کامل

Non-I.I.D. Multi-Instance Dimensionality Reduction by Learning a Maximum Bag Margin Subspace

Multi-instance learning, as other machine learning tasks, also suffers from the curse of dimensionality. Although dimensionality reduction methods have been investigated for many years, multi-instance dimensionality reduction methods remain untouched. On the other hand, most algorithms in multi-instance framework treat instances in each bag as independently and identically distributed (i.i.d.) ...

متن کامل

Large Margin Low Rank Tensor Analysis

We present a supervised model for tensor dimensionality reduction, which is called large margin low rank tensor analysis (LMLRTA). In contrast to traditional vector representation-based dimensionality reduction methods, LMLRTA can take any order of tensors as input. And unlike previous tensor dimensionality reduction methods, which can learn only the low-dimensional embeddings with a priori spe...

متن کامل

Learning Through Non-linearly Supervised Dimensionality Reduction

Dimensionality reduction is a crucial ingredient of machine learning and data mining, boosting classification accuracy through the isolation of patterns via omission of noise. Nevertheless, recent studies have shown that dimensionality reduction can benefit from label information, via a joint estimation of predictors and target variables from a low-rank representation. In the light of such insp...

متن کامل

Large-margin Weakly Supervised Dimensionality Reduction

This paper studies dimensionality reduction in a weakly supervised setting, in which the preference relationship between examples is indicated by weak cues. A novel framework is proposed that integrates two aspects of the large margin principle (angle and distance), which simultaneously encourage angle consistency between preference pairs and maximize the distance between examples in preference...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008